Music Retrieval and Recommendation Scheme Based on Varying Mood Sequences

نویسندگان

  • Sanghoon Jun
  • Seungmin Rho
  • Eenjun Hwang
چکیده

A typical music clip consists of one or more segments with different moods and such mood information could be a crucial clue for determining the similarity between music clips. One representative mood has been selected for music clip for retrieval, recommendation or classification purposes, which often gives unsatisfactory result. In this paper, the authors propose a new music retrieval and recommendation scheme based on the mood sequence of music clips. The authors first divide each music clip into segments through beat structure analysis, then, apply the k-medoids clustering algorithm for grouping all the segments into clusters with similar features. By assigning a unique mood symbol for each cluster, one can transform each music clip into a musical mood sequence. For music retrieval, the authors use the Smith-Waterman (SW) algorithm to measure the similarity between mood sequences. However, for music recommendation, user preferences are retrieved from a recent music playlist or user interaction through the interface, which generates a music recommendation list based on the mood sequence similarity. The authors demonstrate that the proposed scheme achieves excellent performance in terms of retrieval accuracy and user satisfaction in music recommendation. dependent on the media type. For instance, in the case of audio data, low-level acoustic features such as MFCCs and ZCR are extracted from audio clips and indexed for retrieval purposes (Ghias, Logan, Chanberlin, & Smith, 1995; Birmingham, Dannenberg, & Pardo, 2006; Rho, Han, Hwang, & Kim, 2008). However, these low-level features fail to provide semantic information of music contents such as mood or emotion. This is a serious limitation in retrieving and recommending appropriate music in DOI: 10.4018/jswis.2010040101 2 International Journal on Semantic Web and Information Systems, 6(2), 1-16, April-June 2010 Copyright © 2010, IGI Global. Copying or distributing in print or electronic forms without written permission of IGI Global is prohibited. many applications. To handle this limitation, many researchers have tried to utilize high-level musical features such as harmonics and beat. On the other hand, many researchers (Feng, Zhuang, & Pan, 2003; Juslin & Sloboda, 2001; Lu, Liu, & Zhang, 2006) have investigated the influence of music features such as loudness and tonality on the perceived emotional expression. They analyzed such data using diverse techniques, some of which are involved in measuring psychological and physiological correlations between the state of a particular musical factor and emotion evocation. Feng et al. (2003) approached the problem from the viewpoint of Computational Media Aesthetics (CMA) and mapped two dimensions of tempo and articulation into four categories of moods: happiness, anger, sadness and fear. This categorization is based on Juslin’s theory (Juslin & Sloboda, 2001) in which the authors investigated the utilization of acoustic cues in the communication of music emotions between performers and listeners, and measured the correlations between emotional expressions (i.e., anger, sadness and happiness) and acoustic cues (i.e., tempo, spectrum and articulation). Lu et al. (2006) classified various features into three categories: intensity, timber and rhythm and mapped all the moods into Thayer’s twodimensional space (Juslin & Sloboda, 2001). They used the Gaussian mixture model (GMM) as a classifier. To track music moods, they considered musical mood variation and proposed a mood boundary detection scheme based on threshold adaptation. To the best of our knowledge, no effort has been made to measure the similarities of music clips based on the musical mood change pattern. Some music genres such as classical usually invoke more than one musical mood. For instance, Beethoven’s Symphony No. 5 is famous for mood changes ranging from magnificence to nimbleness. Distinct musical features invoke different musical moods. In Jun et al. (2009), we considered a single global feature for mood classification and used it for music retrieval. However, in many cases, a single global feature could not summarize various moods in a music clip effectively and hence provide incorrect result. To solve this problem, in this paper, we first divide music clips into segments, extract various low-level features from each segment and then classify them into clusters with similar features. By assigning an appropriate mood symbol for each cluster, we can get musical mood sequences for music clips and use them for music retrieval and recommendation. The rest of this paper is organized as follows: In Section 2, we present a brief overview on the recent music retrieval and mood classification techniques. Section 3 presents the overall system architecture and some details on the similar music retrieval and recommendation. Section 4 describes the experiments we performed and some of the results. In the last section, we conclude the paper with directions for future work.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Cross-cultural Study of Music Mood Perception between American and Chinese Listeners

Music mood has been recognized as an important access point for music and many online music services support browsing by mood. However, how people judge music mood has not been well studied in the Music Information Retrieval (MIR) domain. In particular, people's cultural background is often assumed to be an important factor in music mood perception, but this assumption has not been verified by ...

متن کامل

Music recommendation for music learning: Hotttabs, a multimedia guitar tutor

Music recommendation systems built on top of music information retrieval (MIR) technologies are usually designed to provide new ways to discover and listen to digital music collections. However, they do not typically assist in another important aspect of musical activity, music learning. In this study we present the application Hotttabs, an online music recommendation system dedicated to guitar...

متن کامل

Predicting High-level Music Semantics Using Social Tags via Ontology-based Reasoning

High-level semantics such as “mood” and “usage” are very useful in music retrieval and recommendation but they are normally hard to acquire. Can we predict them from a cloud of social tags? We propose a semantic identification and reasoning method: Given a music taxonomy system, we map it to an ontology’s terminology, map its finite set of terms to the ontology’s assertional axioms, and then ma...

متن کامل

A Machine Learning based Music Retrieval and Recommendation System

In this paper, we present a music retrieval and recommendation system using machine learning techniques. We propose a query by humming system for music retrieval that uses deep neural networks for note transcription and a note-based retrieval system for retrieving the correct song from the database. We evaluate our query by humming system using the standard MIREX QBSH dataset. We also propose a...

متن کامل

Using Automated Rhyme Detection to Characterize Rhyming Style in Rap Music

Imperfect and internal rhymes are two important features in rap music previously ignored in the music information retrieval literature. We developed a method of scoring potential rhymes using a probabilistic model based on phoneme frequencies in rap lyrics. We used this scoring scheme to automatically identify internal and line-final rhymes in song lyrics and demonstrated the performance of thi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Int. J. Semantic Web Inf. Syst.

دوره 6  شماره 

صفحات  -

تاریخ انتشار 2010